Decreasing the Randomness of Random Forests

نویسندگان

  • Samuel Robert Reid
  • Samuel R. Reid
چکیده

The Random Forest algorithm is an ensemble technique that can achieve high accuracy on classification and regression with minimal tuning of parameters. This paper analyzes the effectiveness of the Random Forest classification algorithm under decreasing randomness in the bootstrap sampling procedure, in increasing tournament size, and in tournament participant selection.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Approaches to Combine Techniques Used by Ensemble Learning Methods ⋆

Discuss approaches to combine techniques used by ensemble learning methods. Randomness which is used by Bagging and Random Forests is introduced into Adaboost to get robust performance under noisy situation. Declare that when the randomness introduced into AdaBoost equals to 100, the proposed algorithm turns out to be a Random Forests with weight update technique. Approaches are discussed to im...

متن کامل

The Utility of Randomness in Decision Tree Ensembles

The use of randomness in constructing decision tree ensembles has drawn much attention in the machine learning community. In general, ensembles introduce randomness to generate diverse trees and in turn they enhance ensembles’ predictive accuracy. Examples of such ensembles are Bagging, Random Forests and Random Decision Tree. In the past, most of the random tree ensembles inject various kinds ...

متن کامل

Applying Randomness Effectively Based on Random Forests for Classification Task of Datasets of Insufficient Information

Random forests are known to be good for data mining of classification tasks, because random forests are robust for datasets having insufficient information possibly with some errors. But applying random forests blindly may not produce good results, and a dataset in the domain of rotogravure printing is one of such datasets. Hence, in this paper, some best classification accuracy based on clever...

متن کامل

Measuring the Algorithmic Convergence of Random Forests via Bootstrap Extrapolation

When making predictions with a voting rule, a basic question arises: “What is the smallest number of votes needed to make a good prediction?” In the context of ensemble classifiers, such as Random Forests or Bagging, this question represents a tradeoff between computational cost and statistical performance. Namely, by paying a larger computational price for more classifiers, the prediction erro...

متن کامل

A Randomness Test for Stable Data

In this paper, we propose a new method for checking randomness of non-Gaussian stable data based on a characterization result. This method is more sensitive with respect to non-random data compared to the well-known non-parametric randomness tests.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004